DE eng

Search in the Catalogues and Directories

Page: 1 2 3 4 5
Hits 1 – 20 of 81

1
MAGIC DUST FOR CROSS-LINGUAL ADAPTATION OF MONOLINGUAL WAV2VEC-2.0
In: ICASSP 2022 ; https://hal.archives-ouvertes.fr/hal-03544515 ; ICASSP 2022, May 2022, Singapour, Singapore (2022)
BASE
Show details
2
Simple and Effective Unsupervised Speech Synthesis ...
BASE
Show details
3
Learning Audio-Video Language Representations
Rouditchenko, Andrew. - : Massachusetts Institute of Technology, 2021
BASE
Show details
4
Cascaded Multilingual Audio-Visual Learning from Videos ...
BASE
Show details
5
Magic dust for cross-lingual adaptation of monolingual wav2vec-2.0 ...
BASE
Show details
6
Text-Free Image-to-Speech Synthesis Using Learned Segmental Units ...
BASE
Show details
7
Exposure Bias versus Self-Recovery: Are Distortions Really Incremental for Autoregressive Text Generation? ...
BASE
Show details
8
Mitigating Biases in Toxic Language Detection through Invariant Rationalization ...
BASE
Show details
9
Mitigating Biases in Toxic Language Detection through Invariant Rationalization ...
BASE
Show details
10
A Convolutional Deep Markov Model for Unsupervised Speech Representation Learning
In: Interspeech 2020 ; https://hal.archives-ouvertes.fr/hal-02912029 ; Interspeech 2020, Oct 2020, Shanghai, China (2020)
BASE
Show details
11
Similarity Analysis of Contextual Word Representation Models ...
BASE
Show details
12
CSTNet: Contrastive Speech Translation Network for Self-Supervised Speech Representation Learning ...
BASE
Show details
13
A Convolutional Deep Markov Model for Unsupervised Speech Representation Learning ...
BASE
Show details
14
What Was Written vs. Who Read It: News Media Profiling Using Text Analysis and Social Media Context ...
BASE
Show details
15
Vector-Quantized Autoregressive Predictive Coding ...
Chung, Yu-An; Tang, Hao; Glass, James. - : arXiv, 2020
BASE
Show details
16
Non-Autoregressive Predictive Coding for Learning Speech Representations from Local Dependencies ...
BASE
Show details
17
Improved Speech Representations with Multi-Target Autoregressive Predictive Coding ...
Chung, Yu-An; Glass, James. - : arXiv, 2020
BASE
Show details
18
Classifying Alzheimer's Disease Using Audio and Text-Based Representations of Speech
In: Frontiers (2020)
BASE
Show details
19
Identification of digital voice biomarkers for cognitive health
In: Explor Med (2020)
BASE
Show details
20
On the Linguistic Representational Power of Neural Machine Translation Models
In: Computational Linguistics, Vol 46, Iss 1, Pp 1-52 (2020) (2020)
Abstract: Despite the recent success of deep neural networks in natural language processing and other spheres of artificial intelligence, their interpretability remains a challenge. We analyze the representations learned by neural machine translation (NMT) models at various levels of granularity and evaluate their quality through relevant extrinsic properties. In particular, we seek answers to the following questions: (i) How accurately is word structure captured within the learned representations, which is an important aspect in translating morphologically rich languages? (ii) Do the representations capture long-range dependencies, and effectively handle syntactically divergent languages? (iii) Do the representations capture lexical semantics? We conduct a thorough investigation along several parameters: (i) Which layers in the architecture capture each of these linguistic phenomena; (ii) How does the choice of translation unit (word, character, or subword unit) impact the linguistic properties captured by the underlying representations? (iii) Do the encoder and decoder learn differently and independently? (iv) Do the representations learned by multilingual NMT models capture the same amount of linguistic information as their bilingual counterparts? Our data-driven, quantitative evaluation illuminates important aspects in NMT models and their ability to capture various linguistic phenomena. We show that deep NMT models trained in an end-to-end fashion, without being provided any direct supervision during the training process, learn a non-trivial amount of linguistic information. Notable findings include the following observations: (i) Word morphology and part-of-speech information are captured at the lower layers of the model; (ii) In contrast, lexical semantics or non-local syntactic and semantic dependencies are better represented at the higher layers of the model; (iii) Representations learned using characters are more informed about word-morphology compared to those learned using subword units; and (iv) Representations learned by multilingual models are richer compared to bilingual models.
Keyword: Computational linguistics. Natural language processing; P98-98.5
URL: https://doi.org/10.1162/coli_a_00367
https://doaj.org/article/0f4a3f344db6432ba02ec4d3a127e34d
BASE
Hide details

Page: 1 2 3 4 5

Catalogues
1
0
6
0
0
0
0
Bibliographies
12
0
0
0
0
0
0
0
1
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
65
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern